Contrastive Divergence in Gaussian Diffusions
نویسنده
چکیده
This letter presents an analysis of the contrastive divergence (CD) learning algorithm when applied to continuous-time linear stochastic neural networks. For this case, powerful techniques exist that allow a detailed analysis of the behavior of CD. The analysis shows that CD converges to maximum likelihood solutions only when the network structure is such that it can match the first moments of the desired distribution. Otherwise, CD can converge to solutions arbitrarily different from the log-likelihood solutions, or they can even diverge. This result suggests the need to improve our theoretical understanding of the conditions under which CD is expected to be well behaved and the conditions under which it may fail. In, addition the results point to practical ideas on how to improve the performance of CD.
منابع مشابه
Gaussian Approximations of Small Noise Diffusions in Kullback-leibler Divergence
Abstract. We study Gaussian approximations to the distribution of a diffusion. The approximations are easy to compute: they are defined by two simple ordinary differential equations for the mean and the covariance. Time correlations can also be computed via solution of a linear stochastic differential equation. We show, using the Kullback-Leibler divergence, that the approximations are accurate...
متن کاملFast Communication Gaussian Approximations of Small Noise Diffusions in Kullback–leibler Divergence∗
We study Gaussian approximations to the distribution of a diffusion. The approximations are easy to compute: they are defined by two simple ordinary differential equations for the mean and the covariance. Time correlations can also be computed via solution of a linear stochastic differential equation. We show, using the Kullback–Leibler divergence, that the approximations are accurate in the sm...
متن کاملAn Analysis of Contrastive Divergence Learning in Gaussian Boltzmann Machines
The Boltzmann machine (BM) learning rule for random field models with latent variables can be problematic to use in practice. These problems have (at least partially) been attributed to the negative phase in BM learning where a Gibbs sampling chain should be run to equilibrium. Hinton (1999, 2000) has introduced an alternative called contrastive divergence (CD) learning where the chain is run f...
متن کاملContrastive Divergence In Gaussian Diffusion Processes
Contrastive Divergence (CD) is a recent learning rule found to work very well in practice despite its still unclear theoretical underpinnings (Carreira-Perpiñan and Hinton, 2005; Hinton, 2002; Hinton and Salakhutdinov, 2006; Hyvärinen, 2006; MacKay, 2001; Roth and Black, 2005; Williams and Agakov, 2002; Yuille, 2004). This paper presents a theoretical analysis of CD in Gaussian Diffusion Proces...
متن کاملDifferential Contrastive Divergence
We formulate a differential version of contrastive divergence for continuous configuration spaces by considering a limit of MCMC processes in which the proposal distribution becomes infinitesimal. This leads to a deterministic differential contrastive divergence update — one in which no stochastic sampling is required. We prove convergence of differential contrastive divergence in general and p...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Neural computation
دوره 20 9 شماره
صفحات -
تاریخ انتشار 2008